Question Answering with Character-Level LSTM Encoders and Model-Based Data Augmentation

نویسندگان

  • Runze Wang
  • Chen-Di Zhan
  • Zhen-Hua Ling
چکیده

This paper presents a character-level encoder-decoder modeling method for question answering (QA) from large-scale knowledge bases (KB). This method improves the existing approach [9] from three aspects. First, long short-term memory (LSTM) structures are adopted to replace the convolutional neural networks (CNN) for encoding the candidate entities and predicates. Second, a new strategy of generating negative samples for model training is adopted. Third, a data augmentation strategy is applied to increase the size of the training set by generating factoid questions using another trained encoder-decoder model. Experimental results on the SimpleQuestions dataset and the Freebase5M KB demonstrates the effectiveness of the proposed method, which improves the state-of-the-art accuracy from 70.3% to 78.8% when augmenting the training set with 70,000 generated triple-question pairs.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

CS224n PA4: Extending Match-LSTM

We propose two novel extensions to the Match-LSTM Boundary model for question answering on the SQuAD dataset. First we propose doing attention in the passage and question encoders. Second we propose adding a one-way conditional dependency between start-of-span and end-of-span prediction. In our evaluations, we show that these extensions result in a model that outperforms our implementation of v...

متن کامل

Character-Level Question Answering with Attention

We show that a character-level encoderdecoder framework can be successfully applied to question answering with a structured knowledge base. We use our model for singlerelation question answering and demonstrate the effectiveness of our approach on the SimpleQuestions dataset (Bordes et al., 2015), where we improve state-of-the-art accuracy from 63.9% to 70.9%, without use of ensembles. Importan...

متن کامل

Deep Poetry: Word-Level and Character-Level Language Models for Shakespearean Sonnet Generation

Text generation is a foundational task in natural language processing, forming the core of a diverse set of practical applications ranging from image captioning and text summarization to question answering. However, most of this work has focused on generating prose. We investigate whether deep learning systems can be used to synthesize poetry, in particular Shakespearean-styled works. Previous ...

متن کامل

Reading Comprehension with Coattention Encoders

We present an end-to-end network architecture for question answering task given a real-world paragraph context. Using SQuAD dataset composed of 100K pairs, we train a coattention encoder to obtain a knowledge representation of the question and context data. This in turn is given as input to the decoder, which generates answer spans within the context. The prediction ...

متن کامل

YNU-HPCC at IJCNLP-2017 Task 5: Multi-choice Question Answering in Exams Using an Attention-based LSTM Model

A shared task is a typical question answering task that aims to test how accurately the participants can answer the questions in exams. Typically, for each question, there are four candidate answers, and only one of the answers is correct. The existing methods for such a task usually implement a recurrent neural network (RNN) or long short-term memory (LSTM). However, both RNN and LSTM are bias...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017